Thesis Covariance Regularization in Mixture of Gaussians for High-dimensional Image Classification
نویسنده
چکیده
OF THESIS COVARIANCE REGULARIZATION IN MIXTURE OF GAUSSIANS FOR HIGH-DIMENSIONAL IMAGE CLASSIFICATION In high dimensions, it is rare to find a data set large enough to compute a non-singular covariance matrix. This problem is exacerbated when performing clustering using a mixture of Gaussians (MoG) because now each cluster’s covariance matrix is computed from only a subset of the data set making the presence of a sufficient amount of data even less likely. The objective of this Thesis is to study the effect of performing a MoG where the covariance matrix is regularized in high-dimensional spaces. Furthermore we present several long-standing algorithms as MoG with various forms of covariance regularization. From this new perspective of MoG with covariance regularization, we begin taking a look at the effect of covariance regularization on MoG classification accuracy. Our secondary objective is to use our results to begin a discussion about the merits of using increased amounts of covariance for MoG on high-dimensional data. We experiment with three different covariance regularization methods What we find is that the simplest form of covariance regularization experimented with in this Thesis results in the best classification of our natural image data sets. Our experiments clearly show that, as the dimensionality of the data decreases, so does the desired or necessary level of covariance regularization. Daniel L Elliott Department of Computer Science Colorado State University Fort Collins, Colorado 80523 Spring 2009
منابع مشابه
Covariance Regularization for Supervised Learning in High Dimensions
This paper studies the effect of covariance regularization for classification of high-dimensional data. This is done by fitting a mixture of Gaussians with a regularized covariance matrix to each class. Three data sets are chosen to suggest the results are applicable to any domain with high-dimensional data. The regularization needs of the data when pre-processed using the dimensionality reduct...
متن کاملCS 545 : Assignment 8 Dan
Mixture of Probabilistic Principal Component Analyzers (MPPCA) is a seminal work in Machine Learning in that it was the first to use PCA to perform clustering and local dimensionality reduction. MPPCA is based upon the mixture of Factor Analyzers (MFA) which is similar to MPPCA except is uses Factor Analysis to estimate the covariance matrix. This algorithm is of interest to me because it is re...
متن کاملImage Segmentation using Gaussian Mixture Model
Abstract: Stochastic models such as mixture models, graphical models, Markov random fields and hidden Markov models have key role in probabilistic data analysis. In this paper, we used Gaussian mixture model to the pixels of an image. The parameters of the model were estimated by EM-algorithm. In addition pixel labeling corresponded to each pixel of true image was made by Bayes rule. In fact,...
متن کاملIMAGE SEGMENTATION USING GAUSSIAN MIXTURE MODEL
Stochastic models such as mixture models, graphical models, Markov random fields and hidden Markov models have key role in probabilistic data analysis. In this paper, we have learned Gaussian mixture model to the pixels of an image. The parameters of the model have estimated by EM-algorithm. In addition pixel labeling corresponded to each pixel of true image is made by Bayes rule. In fact, ...
متن کاملGaussian mixture models for the classification of high-dimensional vibrational spectroscopy data
In this work, a family of generative Gaussian models designed for the supervised classification of high-dimensional data is presented as well as the associated classification method called High Dimensional Discriminant Analysis (HDDA). The features of these Gaussian models are: i) the representation of the input density model is smooth; ii) the data of each class are modeled in a specific subsp...
متن کامل